Your browser doesn't support javascript.
Show: 20 | 50 | 100
Results 1 - 13 de 13
Filter
Add filters

Document Type
Year range
1.
Velma Lopez; Estee Y Cramer; Robert Pagano; John M Drake; Eamon B O'Dea; Benjamin P Linas; Turgay Ayer; Jade Xiao; Madeline Adee; Jagpreet Chhatwal; Mary A Ladd; Peter P Mueller; Ozden O Dalgic; Johannes Bracher; Tilmann Gneiting; Anja Mühlemann; Jarad Niemi; Ray L Evan; Martha Zorn; Yuxin Huang; Yijin Wang; Aaron Gerding; Ariane Stark; Dasuni Jayawardena; Khoa Le; Nutcha Wattanachit; Abdul H Kanji; Alvaro J Castro Rivadeneira; Sen Pei; Jeffrey Shaman; Teresa K Yamana; Xinyi Li; Guannan Wang; Lei Gao; Zhiling Gu; Myungjin Kim; Lily Wang; Yueying Wang; Shan Yu; Daniel J Wilson; Samuel R Tarasewicz; Brad Suchoski; Steve Stage; Heidi Gurung; Sid Baccam; Maximilian Marshall; Lauren Gardner; Sonia Jindal; Kristen Nixon; Joseph C Lemaitre; Juan Dent; Alison L Hill; Joshua Kaminsky; Elizabeth C Lee; Justin Lessler; Claire P Smith; Shaun Truelove; Matt Kinsey; Katharine Tallaksen; Shelby Wilson; Luke C Mullany; Lauren Shin; Kaitlin Rainwater-Lovett; Dean Karlen; Lauren Castro; Geoffrey Fairchild; Isaac Michaud; Dave Osthus; Alessandro Vespignani; Matteo Chinazzi; Jessica T Davis; Kunpeng Mu; Xinyue Xiong; Ana Pastore y Piontti; Shun Zheng; Zhifeng Gao; Wei Cao; Jiang Bian; Chaozhuo Li; Xing Xie; Tie-Yan Liu; Juan Lavista Ferres; Shun Zhang; Robert Walraven; Jinghui Chen; Quanquan Gu; Lingxiao Wang; Pan Xu; Weitong Zhang; Difan Zou; Graham Casey Gibson; Daniel Sheldon; Ajitesh Srivastava; Aniruddha Adiga; Benjamin Hurt; Gursharn Kaur; Bryan Lewis; Madhav Marathe; Akhil S Peddireddy; Przemyslaw Porebski; Srinivasan Venkatramanan; Lijing Wang; Pragati V Prasad; Alexander E Webber; Jo W Walker; Rachel B Slayton; Matthew Biggerstaff; Nicholas G Reich; Michael A Johansson.
medrxiv; 2023.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2023.05.30.23290732

ABSTRACT

During the COVID-19 pandemic, forecasting COVID-19 trends to support planning and response was a priority for scientists and decision makers alike. In the United States, COVID-19 forecasting was coordinated by a large group of universities, companies, and government entities led by the Centers for Disease Control and Prevention and the US COVID-19 Forecast Hub (https://covid19forecasthub.org). We evaluated approximately 9.7 million forecasts of weekly state-level COVID-19 cases for predictions 1-4 weeks into the future submitted by 24 teams from August 2020 to December 2021. We assessed coverage of central prediction intervals and weighted interval scores (WIS), adjusting for missing forecasts relative to a baseline forecast, and used a Gaussian generalized estimating equation (GEE) model to evaluate differences in skill across epidemic phases that were defined by the effective reproduction number. Overall, we found high variation in skill across individual models, with ensemble-based forecasts outperforming other approaches. Forecast skill relative to the baseline was generally higher for larger jurisdictions (e.g., states compared to counties). Over time, forecasts generally performed worst in periods of rapid changes in reported cases (either in increasing or decreasing epidemic phases) with 95% prediction interval coverage dropping below 50% during the growth phases of the winter 2020, Delta, and Omicron waves. Ideally, case forecasts could serve as a leading indicator of changes in transmission dynamics. However, while most COVID-19 case forecasts outperformed a naive baseline model, even the most accurate case forecasts were unreliable in key phases. Further research could improve forecasts of leading indicators, like COVID-19 cases, by leveraging additional real-time data, addressing performance across phases, improving the characterization of forecast confidence, and ensuring that forecasts were coherent across spatial scales. In the meantime, it is critical for forecast users to appreciate current limitations and use a broad set of indicators to inform pandemic-related decision making. Author SummaryAs SARS-CoV-2 began to spread throughout the world in early 2020, modelers played a critical role in predicting how the epidemic could take shape. Short-term forecasts of epidemic outcomes (for example, infections, cases, hospitalizations, or deaths) provided useful information to support pandemic planning, resource allocation, and intervention. Yet, infectious disease forecasting is still a nascent science, and the reliability of different types of forecasts is unclear. We retrospectively evaluated COVID-19 case forecasts, which were often unreliable. For example, forecasts did not anticipate the speed of increase in cases in early winter 2020. This analysis provides insights on specific problems that could be addressed in future research to improve forecasts and their use. Identifying the strengths and weaknesses of forecasts is critical to improving forecasting for current and future public health responses.


Subject(s)
COVID-19 , Death , Communicable Diseases
2.
medrxiv; 2023.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2023.04.27.23289109

ABSTRACT

The effective reproductive number Rt has taken a central role in the scientific, political, and public discussion during the COVID-19 pandemic, with numerous real-time estimates of this quantity routinely published. Disagreement between estimates can be substantial and may lead to confusion among decision-makers and the general public. In this work, we compare different estimates of the national-level effective reproductive number of COVID-19 in Germany in 2020 and 2021. We consider the agreement between estimates from the same method but published at different time points (within-method agreement) as well as retrospective agreement across different approaches (between-method agreement). Concerning the former, estimates from some methods are very stable over time and hardly subject to revisions, while others display considerable fluctuations. To evaluate between-method agreement, we reproduce the estimates generated by different groups using a variety of statistical approaches, standardizing analytical choices to assess how they contribute to the observed disagreement. These analytical choices include the data source, data pre-processing, assumed generation time distribution, statistical tuning parameters, and various delay distributions. We find that in practice, these auxiliary choices in the estimation of Rt may affect results at least as strongly as the selection of the statistical approach. They should thus be communicated transparently along with the estimates.


Subject(s)
COVID-19 , Confusion
3.
medrxiv; 2023.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2023.04.17.23288668

ABSTRACT

Real-time surveillance is a crucial element in the response to infectious disease outbreaks. However, the interpretation of incidence data is often hampered by delays occurring at various stages of data gathering and reporting. As a result, recent values are biased downward, which obscures current trends. Statistical nowcasting techniques can be employed to correct these biases, allowing for accurate characterization of recent developments and thus enhancing situational awareness. In this paper, we present a preregistered real-time assessment of eight nowcasting approaches, applied by independent research teams to German 7-day hospitalization incidences. This indicator played an important role in the management of the pandemic in Germany and was linked to levels of non-pharmaceutical interventions via certain thresholds. Due to its definition, in which hospitalization counts are aggregated by the date of case report rather than admission, German hospitalization incidences are particularly affected by delays and can take several weeks or months to fully stabilize. For this study, all methods were applied from 22 November 2021 to 29 April 2022, with probabilistic nowcasts produced each day for the current and 28 preceding days. Nowcasts at the national, state, and age-group levels were collected in the form of quantiles in a public repository and displayed in a dashboard. Moreover, a mean and a median ensemble nowcast were generated. We find that overall, the compared methods were able to remove a large part of the biases introduced by delays. Most participating teams underestimated the importance of very long delays, though, resulting in nowcasts with a slight downward bias. The accompanying uncertainty intervals were also too narrow for almost all methods. Averaged over all nowcast horizons, the best performance was achieved by a model using case incidences as a covariate and taking into account longer delays than the other approaches. For the most recent days, which are often considered the most relevant in practice, a mean ensemble of the submitted nowcasts performed best. We conclude by providing some lessons learned on the definition of nowcasting targets and practical challenges.


Subject(s)
COVID-19 , Communicable Diseases
4.
medrxiv; 2022.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2022.10.12.22280917

ABSTRACT

Multi-model and multi-team ensemble forecasts have become widely used to generate reliable short-term predictions of infectious disease spread. Notably, various public health agencies have used them to leverage academic disease modelling during the COVID-19 pandemic. However, ensemble forecasts are difficult to interpret and require extensive effort from numerous participating groups as well as a coordination team. In other fields, resource usage has been reduced by training simplified models that reproduce some of the observed behaviour of more complex models. Here we used observations of the behaviour of the European COVID-19 Forecast Hub ensemble combined with our own forecasting experience to identify a set of properties present in current ensemble forecasts. We then developed a parsimonious forecast model intending to mirror these properties. We assess forecasts generated from this model in real time over six months (the 15th of January 2022 to the 19th o July 2022) and for multiple European countries. We focused on forecasts of cases one to four weeks ahead and compared them to those by the European forecast hub ensemble. We find that the surrogate model behaves qualitatively similarly to the ensemble in many instances, though with increased uncertainty and poorer performance around periods of peak incidence (as measured by the Weighted Interval Score). The performance differences, however, seem to be partially due to a subset of time points, and the proposed model appears better probabilistically calibrated than the ensemble. We conclude that our simplified forecast model may have captured some of the dynamics of the hub ensemble, but more work is needed to understand the implicit epidemiological model that it represents.


Subject(s)
COVID-19 , Learning Disabilities , Communicable Diseases
5.
Katharine Sherratt; Hugo Gruson; Rok Grah; Helen Johnson; Rene Niehus; Bastian Prasse; Frank Sandman; Jannik Deuschel; Daniel Wolffram; Sam Abbott; Alexander Ullrich; Graham Gibson; Evan L Ray; Nicholas G Reich; Daniel Sheldon; Yijin Wang; Nutcha Wattanachit; Lijing Wang; Jan Trnka; Guillaume Obozinski; Tao Sun; Dorina Thanou; Loic Pottier; Ekaterina Krymova; Maria Vittoria Barbarossa; Neele Leithauser; Jan Mohring; Johanna Schneider; Jaroslaw Wlazlo; Jan Fuhrmann; Berit Lange; Isti Rodiah; Prasith Baccam; Heidi Gurung; Steven Stage; Bradley Suchoski; Jozef Budzinski; Robert Walraven; Inmaculada Villanueva; Vit Tucek; Martin Smid; Milan Zajicek; Cesar Perez Alvarez; Borja Reina; Nikos I Bosse; Sophie Meakin; Pierfrancesco Alaimo Di Loro; Antonello Maruotti; Veronika Eclerova; Andrea Kraus; David Kraus; Lenka Pribylova; Bertsimas Dimitris; Michael Lingzhi Li; Soni Saksham; Jonas Dehning; Sebastian Mohr; Viola Priesemann; Grzegorz Redlarski; Benjamin Bejar; Giovanni Ardenghi; Nicola Parolini; Giovanni Ziarelli; Wolfgang Bock; Stefan Heyder; Thomas Hotz; David E. Singh; Miguel Guzman-Merino; Jose L Aznarte; David Morina; Sergio Alonso; Enric Alvarez; Daniel Lopez; Clara Prats; Jan Pablo Burgard; Arne Rodloff; Tom Zimmermann; Alexander Kuhlmann; Janez Zibert; Fulvia Pennoni; Fabio Divino; Marti Catala; Gianfranco Lovison; Paolo Giudici; Barbara Tarantino; Francesco Bartolucci; Giovanna Jona Lasinio; Marco Mingione; Alessio Farcomeni; Ajitesh Srivastava; Pablo Montero-Manso; Aniruddha Adiga; Benjamin Hurt; Bryan Lewis; Madhav Marathe; Przemyslaw Porebski; Srinivasan Venkatramanan; Rafal Bartczuk; Filip Dreger; Anna Gambin; Krzysztof Gogolewski; Magdalena Gruziel-Slomka; Bartosz Krupa; Antoni Moszynski; Karol Niedzielewski; Jedrzej Nowosielski; Maciej Radwan; Franciszek Rakowski; Marcin Semeniuk; Ewa Szczurek; Jakub Zielinski; Jan Kisielewski; Barbara Pabjan; Kirsten Holger; Yuri Kheifetz; Markus Scholz; Marcin Bodych; Maciej Filinski; Radoslaw Idzikowski; Tyll Krueger; Tomasz Ozanski; Johannes Bracher; Sebastian Funk.
medrxiv; 2022.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2022.06.16.22276024

ABSTRACT

Background: Short-term forecasts of infectious disease burden can contribute to situational awareness and aid capacity planning. Based on best practice in other fields and recent insights in infectious disease epidemiology, one can maximise the predictive performance of such forecasts if multiple models are combined into an ensemble. Here we report on the performance of ensembles in predicting COVID-19 cases and deaths across Europe between 08 March 2021 and 07 March 2022. Methods: We used open-source tools to develop a public European COVID-19 Forecast Hub. We invited groups globally to contribute weekly forecasts for COVID-19 cases and deaths reported from a standardised source over the next one to four weeks. Teams submitted forecasts from March 2021 using standardised quantiles of the predictive distribution. Each week we created an ensemble forecast, where each predictive quantile was calculated as the equally-weighted average (initially the mean and then from 26th July the median) of all individual models predictive quantiles. We measured the performance of each model using the relative Weighted Interval Score (WIS), comparing models forecast accuracy relative to all other models. We retrospectively explored alternative methods for ensemble forecasts, including weighted averages based on models past predictive performance. Results: Over 52 weeks we collected and combined up to 28 forecast models for 32 countries. We found a weekly ensemble had a consistently strong performance across countries over time. Across all horizons and locations, the ensemble performed better on relative WIS than 84% of participating models forecasts of incident cases (with a total N=862), and 92% of participating models forecasts of deaths (N=746). Across a one to four week time horizon, ensemble performance declined with longer forecast periods when forecasting cases, but remained stable over four weeks for incident death forecasts. In every forecast across 32 countries, the ensemble outperformed most contributing models when forecasting either cases or deaths, frequently outperforming all of its individual component models. Among several choices of ensemble methods we found that the most influential and best choice was to use a median average of models instead of using the mean, regardless of methods of weighting component forecast models. Conclusions: Our results support the use of combining forecasts from individual models into an ensemble in order to improve predictive performance across epidemiological targets and populations during infectious disease epidemics. Our findings further suggest that median ensemble methods yield better predictive performance more than ones based on means. Our findings also highlight that forecast consumers should place more weight on incident death forecasts than incident case forecasts at forecast horizons greater than two weeks.


Subject(s)
COVID-19 , Death , Communicable Diseases
6.
arxiv; 2022.
Preprint in English | PREPRINT-ARXIV | ID: ppzbmed-2201.12387v2

ABSTRACT

The U.S. COVID-19 Forecast Hub aggregates forecasts of the short-term burden of COVID-19 in the United States from many contributing teams. We study methods for building an ensemble that combines forecasts from these teams. These experiments have informed the ensemble methods used by the Hub. To be most useful to policy makers, ensemble forecasts must have stable performance in the presence of two key characteristics of the component forecasts: (1) occasional misalignment with the reported data, and (2) instability in the relative performance of component forecasters over time. Our results indicate that in the presence of these challenges, an untrained and robust approach to ensembling using an equally weighted median of all component forecasts is a good choice to support public health decision makers. In settings where some contributing forecasters have a stable record of good performance, trained ensembles that give those forecasters higher weight can also be helpful.


Subject(s)
COVID-19
7.
medrxiv; 2021.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2021.12.01.21266598

ABSTRACT

Forecasts based on epidemiological modelling have played an important role in shaping public policy throughout the COVID-19 pandemic. This modelling combines knowledge about infectious disease dynamics with the subjective opinion of the researcher who develops and refines the model and often also adjusts model outputs. Developing a forecast model is difficult, resource- and time-consuming. It is therefore worth asking what modelling is able to add beyond the subjective opinion of the researcher alone. To investigate this, we analysed different real-time forecasts of cases of and deaths from COVID-19 in Germany and Poland over a 1-4 week horizon submitted to the German and Polish Forecast Hub. We compared crowd forecasts elicited from researchers and volunteers, against a) forecasts from two semi-mechanistic models based on common epidemiological assumptions and b) the ensemble of all other models submitted to the Forecast Hub. We found crowd forecasts, despite being overconfident, to outperform all other methods across all forecast horizons when forecasting cases (weighted interval score relative to the Hub ensemble 2 weeks ahead: 0.89). Forecasts based on computational models performed comparably better when predicting deaths (rel. WIS 1.26), suggesting that epidemiological modelling and human judgement can complement each other in important ways.


Subject(s)
COVID-19
8.
medrxiv; 2021.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2021.11.05.21265810

ABSTRACT

We report on the second and final part of a pre-registered forecasting study on COVID-19 cases and deaths in Germany and Poland. Fifteen independent research teams provided forecasts at lead times of one through four weeks from January through mid-April 2021. Compared to the first part (October--December 2020), the number of participating teams increased, and a number of teams started providing subnational-level forecasts. The addressed time period is characterized by rather stable non-pharmaceutical interventions in both countries, making short-term predictions more straightforward than in the first part of our study. In both countries, case counts declined initially, before rebounding due to the rise of the B.1.1.7 variant. Deaths declined through most of the study period in Germany while in Poland they increased after a prolonged plateau. Many, though not all, models outperformed a simple baseline model up to four weeks ahead, with ensemble methods showing very good relative performance. Major trend changes in reported cases, however, remained challenging to predict.


Subject(s)
COVID-19
9.
medrxiv; 2021.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2021.11.04.21265886

ABSTRACT

Academic researchers, government agencies, industry groups, and individuals have produced forecasts at an unprecedented scale during the COVID-19 pandemic. To leverage these forecasts, the United States Centers for Disease Control and Prevention (CDC) partnered with an academic research lab at the University of Massachusetts Amherst to create the US COVID-19 Forecast Hub. Launched in April 2020, the Forecast Hub is a dataset with point and probabilistic forecasts of incident hospitalizations, incident cases, incident deaths, and cumulative deaths due to COVID-19 at national, state, and county levels in the United States. Included forecasts represent a variety of modeling approaches, data sources, and assumptions regarding the spread of COVID-19. The goal of this dataset is to establish a standardized and comparable set of short-term forecasts from modeling teams. These data can be used to develop ensemble models, communicate forecasts to the public, create visualizations, compare models, and inform policies regarding COVID-19 mitigation. These open-source data are available via download from GitHub, through an online API, and through R packages.


Subject(s)
COVID-19
10.
Estee Y Cramer; Evan L Ray; Velma K Lopez; Johannes Bracher; Andrea Brennen; Alvaro J Castro Rivadeneira; Aaron Gerding; Tilmann Gneiting; Katie H House; Yuxin Huang; Dasuni Jayawardena; Abdul H Kanji; Ayush Khandelwal; Khoa Le; Anja Muhlemann; Jarad Niemi; Apurv Shah; Ariane Stark; Yijin Wang; Nutcha Wattanachit; Martha W Zorn; Youyang Gu; Sansiddh Jain; Nayana Bannur; Ayush Deva; Mihir Kulkarni; Srujana Merugu; Alpan Raval; Siddhant Shingi; Avtansh Tiwari; Jerome White; Spencer Woody; Maytal Dahan; Spencer Fox; Kelly Gaither; Michael Lachmann; Lauren Ancel Meyers; James G Scott; Mauricio Tec; Ajitesh Srivastava; Glover E George; Jeffrey C Cegan; Ian D Dettwiller; William P England; Matthew W Farthing; Robert H Hunter; Brandon Lafferty; Igor Linkov; Michael L Mayo; Matthew D Parno; Michael A Rowland; Benjamin D Trump; Sabrina M Corsetti; Thomas M Baer; Marisa C Eisenberg; Karl Falb; Yitao Huang; Emily T Martin; Ella McCauley; Robert L Myers; Tom Schwarz; Daniel Sheldon; Graham Casey Gibson; Rose Yu; Liyao Gao; Yian Ma; Dongxia Wu; Xifeng Yan; Xiaoyong Jin; Yu-Xiang Wang; YangQuan Chen; Lihong Guo; Yanting Zhao; Quanquan Gu; Jinghui Chen; Lingxiao Wang; Pan Xu; Weitong Zhang; Difan Zou; Hannah Biegel; Joceline Lega; Timothy L Snyder; Davison D Wilson; Steve McConnell; Yunfeng Shi; Xuegang Ban; Robert Walraven; Qi-Jun Hong; Stanley Kong; James A Turtle; Michal Ben-Nun; Pete Riley; Steven Riley; Ugur Koyluoglu; David DesRoches; Bruce Hamory; Christina Kyriakides; Helen Leis; John Milliken; Michael Moloney; James Morgan; Gokce Ozcan; Chris Schrader; Elizabeth Shakhnovich; Daniel Siegel; Ryan Spatz; Chris Stiefeling; Barrie Wilkinson; Alexander Wong; Sean Cavany; Guido Espana; Sean Moore; Rachel Oidtman; Alex Perkins; Zhifeng Gao; Jiang Bian; Wei Cao; Juan Lavista Ferres; Chaozhuo Li; Tie-Yan Liu; Xing Xie; Shun Zhang; Shun Zheng; Alessandro Vespignani; Matteo Chinazzi; Jessica T Davis; Kunpeng Mu; Ana Pastore y Piontti; Xinyue Xiong; Andrew Zheng; Jackie Baek; Vivek Farias; Andreea Georgescu; Retsef Levi; Deeksha Sinha; Joshua Wilde; Nicolas D Penna; Leo A Celi; Saketh Sundar; Dave Osthus; Lauren Castro; Geoffrey Fairchild; Isaac Michaud; Dean Karlen; Elizabeth C Lee; Juan Dent; Kyra H Grantz; Joshua Kaminsky; Kathryn Kaminsky; Lindsay T Keegan; Stephen A Lauer; Joseph C Lemaitre; Justin Lessler; Hannah R Meredith; Javier Perez-Saez; Sam Shah; Claire P Smith; Shaun A Truelove; Josh Wills; Matt Kinsey; RF Obrecht; Katharine Tallaksen; John C. Burant; Lily Wang; Lei Gao; Zhiling Gu; Myungjin Kim; Xinyi Li; Guannan Wang; Yueying Wang; Shan Yu; Robert C Reiner; Ryan Barber; Emmanuela Gaikedu; Simon Hay; Steve Lim; Chris Murray; David Pigott; B. Aditya Prakash; Bijaya Adhikari; Jiaming Cui; Alexander Rodriguez; Anika Tabassum; Jiajia Xie; Pinar Keskinocak; John Asplund; Arden Baxter; Buse Eylul Oruc; Nicoleta Serban; Sercan O Arik; Mike Dusenberry; Arkady Epshteyn; Elli Kanal; Long T Le; Chun-Liang Li; Tomas Pfister; Dario Sava; Rajarishi Sinha; Thomas Tsai; Nate Yoder; Jinsung Yoon; Leyou Zhang; Sam Abbott; Nikos I I Bosse; Sebastian Funk; Joel Hellewell; Sophie R Meakin; James D Munday; Katharine Sherratt; Mingyuan Zhou; Rahi Kalantari; Teresa K Yamana; Sen Pei; Jeffrey Shaman; Turgay Ayer; Madeline Adee; Jagpreet Chhatwal; Ozden O Dalgic; Mary A Ladd; Benjamin P Linas; Peter Mueller; Jade Xiao; Michael L Li; Dimitris Bertsimas; Omar Skali Lami; Saksham Soni; Hamza Tazi Bouardi; Yuanjia Wang; Qinxia Wang; Shanghong Xie; Donglin Zeng; Alden Green; Jacob Bien; Addison J Hu; Maria Jahja; Balasubramanian Narasimhan; Samyak Rajanala; Aaron Rumack; Noah Simon; Ryan Tibshirani; Rob Tibshirani; Valerie Ventura; Larry Wasserman; Eamon B O'Dea; John M Drake; Robert Pagano; Jo W Walker; Rachel B Slayton; Michael Johansson; Matthew Biggerstaff; Nicholas G Reich.
medrxiv; 2021.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2021.02.03.21250974

ABSTRACT

Short-term probabilistic forecasts of the trajectory of the COVID-19 pandemic in the United States have served as a visible and important communication channel between the scientific modeling community and both the general public and decision-makers. Forecasting models provide specific, quantitative, and evaluable predictions that inform short-term decisions such as healthcare staffing needs, school closures, and allocation of medical supplies. In 2020, the COVID-19 Forecast Hub (https://covid19forecasthub.org/) collected, disseminated, and synthesized hundreds of thousands of specific predictions from more than 50 different academic, industry, and independent research groups. This manuscript systematically evaluates 23 models that regularly submitted forecasts of reported weekly incident COVID-19 mortality counts in the US at the state and national level. One of these models was a multi-model ensemble that combined all available forecasts each week. The performance of individual models showed high variability across time, geospatial units, and forecast horizons. Half of the models evaluated showed better accuracy than a naive baseline model. In combining the forecasts from all teams, the ensemble showed the best overall probabilistic accuracy of any model. Forecast accuracy degraded as models made predictions farther into the future, with probabilistic accuracy at a 20-week horizon more than 5 times worse than when predicting at a 1-week horizon. This project underscores the role that collaboration and active coordination between governmental public health agencies, academic modeling teams, and industry partners can play in developing modern modeling capabilities to support local, state, and federal response to outbreaks. f


Subject(s)
COVID-19
11.
medrxiv; 2020.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2020.12.24.20248826

ABSTRACT

We report insights from ten weeks of collaborative COVID-19 forecasting for Germany and Poland (12 October - 19 December 2020). The study period covers the onset of the second wave in both countries, with tightening non-pharmaceutical interventions (NPIs) and subsequently a decay (Poland) or plateau and renewed increase (Germany) in reported cases. Thirteen independent teams provided probabilistic real-time forecasts of COVID-19 cases and deaths. These were reported for lead times of one to four weeks, with evaluation focused on one- and two-week horizons, which are less affected by changing NPIs. Heterogeneity between forecasts was considerable both in terms of point predictions and forecast spread. Ensemble forecasts showed good relative performance, in particular in terms of coverage, but did not clearly dominate single-model predictions. The study was preregistered and will be followed up in future phases of the pandemic.


Subject(s)
COVID-19
12.
medrxiv; 2020.
Preprint in English | medRxiv | ID: ppzbmed-10.1101.2020.08.19.20177493

ABSTRACT

Background The COVID-19 pandemic has driven demand for forecasts to guide policy and planning. Previous research has suggested that combining forecasts from multiple models into a single "ensemble" forecast can increase the robustness of forecasts. Here we evaluate the real-time application of an open, collaborative ensemble to forecast deaths attributable to COVID-19 in the U.S. Methods Beginning on April 13, 2020, we collected and combined one- to four-week ahead forecasts of cumulative deaths for U.S. jurisdictions in standardized, probabilistic formats to generate real-time, publicly available ensemble forecasts. We evaluated the point prediction accuracy and calibration of these forecasts compared to reported deaths. Results Analysis of 2,512 ensemble forecasts made April 27 to July 20 with outcomes observed in the weeks ending May 23 through July 25, 2020 revealed precise short-term forecasts, with accuracy deteriorating at longer prediction horizons of up to four weeks. At all prediction horizons, the prediction intervals were well calibrated with 92-96% of observations falling within the rounded 95% prediction intervals. Conclusions This analysis demonstrates that real-time, publicly available ensemble forecasts issued in April-July 2020 provided robust short-term predictions of reported COVID-19 deaths in the United States. With the ongoing need for forecasts of impacts and resource needs for the COVID-19 response, the results underscore the importance of combining multiple probabilistic models and assessing forecast skill at different prediction horizons. Careful development, assessment, and communication of ensemble forecasts can provide reliable insight to public health decision makers.


Subject(s)
COVID-19 , Death
13.
arxiv; 2020.
Preprint in English | PREPRINT-ARXIV | ID: ppzbmed-2005.12881v3

ABSTRACT

For practical reasons, many forecasts of case, hospitalization and death counts in the context of the current COVID-19 pandemic are issued in the form of central predictive intervals at various levels. This is also the case for the forecasts collected in the COVID-19 Forecast Hub (https://covid19forecasthub.org/). Forecast evaluation metrics like the logarithmic score, which has been applied in several infectious disease forecasting challenges, are then not available as they require full predictive distributions. This article provides an overview of how established methods for the evaluation of quantile and interval forecasts can be applied to epidemic forecasts in this format. Specifically, we discuss the computation and interpretation of the weighted interval score, which is a proper score that approximates the continuous ranked probability score. It can be interpreted as a generalization of the absolute error to probabilistic forecasts and allows for a decomposition into a measure of sharpness and penalties for over- and underprediction.


Subject(s)
COVID-19 , Death
SELECTION OF CITATIONS
SEARCH DETAIL